Gaussian Model Selection with an Unknown Variance by Yannick Baraud, Christophe Giraud
نویسنده
چکیده
Let Y be a Gaussian vector whose components are independent with a common unknown variance. We consider the problem of estimating the mean μ of Y by model selection. More precisely, we start with a collection S = {Sm,m ∈ M} of linear subspaces of Rn and associate to each of these the least-squares estimator of μ on Sm. Then, we use a data driven penalized criterion in order to select one estimator among these. Our first objective is to analyze the performance of estimators associated to classical criteria such as FPE, AIC, BIC and AMDL. Our second objective is to propose better penalties that are versatile enough to take into account both the complexity of the collection S and the sample size. Then we apply those to solve various statistical problems such as variable selection, change point detections and signal estimation among others. Our results are based on a nonasymptotic risk bound with respect to the Euclidean loss for the selected estimator. Some analogous results are also established for the Kullback loss.
منابع مشابه
ar X iv : m at h / 07 01 25 0 v 1 [ m at h . ST ] 9 J an 2 00 7 GAUSSIAN MODEL SELECTION WITH UNKNOWN VARIANCE
Let Y be a Gaussian vector whose components are independent with a common unknown variance. We consider the problem of estimating the mean µ of Y by model selection. More precisely, we start with a collection S = {Sm, m ∈ M} of linear subspaces of R n and associate to each of these the least-squares estimator of µ on Sm. Then, we use a data driven penalized criterion in order to select one esti...
متن کاملGaussian Model Selection with an Unknown Variance
Let Y be a Gaussian vector whose components are independent with a common unknown variance. We consider the problem of estimating the mean μ of Y by model selection. More precisely, we start with a collection S = {Sm, m ∈M} of linear subspaces of R and associate to each of these the least-squares estimator of μ on Sm. Then, we use a data driven penalized criterion in order to select one estimat...
متن کاملEstimator selection in the Gaussian setting
We consider the problem of estimating the mean f of a Gaussian vector Y with independent components of common unknown variance σ. Our estimation procedure is based on estimator selection. More precisely, we start with an arbitrary and possibly infinite collection F of estimators of f based on Y and, with the same data Y , aim at selecting an estimator among F with the smallest Euclidean risk. N...
متن کاملGaussian Model Selection with Unknown Variance
Let Y be a Gaussian vector whose components are independent with a common unknown variance. We consider the problem of estimating the mean of Y by model selection. More precisely, we start with a collection S = {Sm, m ∈M} of linear subspaces of R and associate to each of these the least-squares estimator of μ on Sm. Then, we use a data driven penalized criterion in order to select one estimator...
متن کاملConfidence Balls in Gaussian Regression
where f = (f1, . . . , fn) ′ is an unknown vector, σ a positive number and ε1, . . . , εn a sequence of i.i.d. standard Gaussian random variables. For some β ∈ ]0,1[, the aim of this paper is to build a nonasymptotic Euclidean confidence ball for f with probability of coverage 1− β from the observation of Y = (Y1, . . . , Yn) ′. This statistical model includes, as a particular case, the functio...
متن کامل